2024-09-24 00:19:13,305 [ 189846 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/_temp/test/git-repo-copy (runner:42, check_args_and_update_paths) 2024-09-24 00:19:13,306 [ 189846 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/_temp/test/git-repo-copy/tests/integration (runner:90, check_args_and_update_paths) 2024-09-24 00:19:13,306 [ 189846 ] INFO : src dir is not set. Will use /home/ubuntu/_work/_temp/test/git-repo-copy/src (runner:97, check_args_and_update_paths) 2024-09-24 00:19:13,306 [ 189846 ] INFO : base_configs_dir: /home/ubuntu/_work/_temp/test/git-repo-copy/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/_temp/test/git-repo-copy/tests/integration (runner:99, check_args_and_update_paths) 2024-09-24 00:19:13,306 [ 189846 ] INFO : Unknown image altinityinfra/dotnet-client (runner:357, ) Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_o9qf3p --privileged --dns-search='.' --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/_temp/test/git-repo-copy/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/_temp/test/git-repo-copy/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/_temp/test/git-repo-copy/src/Server/grpc_protos:/ClickHouse/src/Server/grpc_protos --volume=/run:/run/host:ro --mount type=bind,source=/home/ubuntu/_work/_temp/test/dockerd_volume_dir,target=/var/lib/docker -e DOCKER_HELPER_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_BASE_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_KERBERIZED_HADOOP_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_KERBEROS_KDC_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_MYSQL_JAVA_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_MYSQL_JS_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_MYSQL_PHP_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_NGINX_DAV_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e CLICKHOUSE_USE_NEW_ANALYZER=1 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 2 -rfEps --run-id=2 --color=no --durations=0 test_mysql_database_engine/test.py::test_mysql_ddl_for_mysql_database test_parallel_replicas_skip_shards/test.py::test_skip_unavailable_shards -vvv" altinityinfra/integration-tests-runner:0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.2.0, repeat-0.9.3, order-1.0.1, xdist-3.5.0, random-0.2 timeout: 900.0s timeout method: signal timeout func_only: False created: 2/2 workers 2 workers [2 items] scheduling tests via LoadFileScheduling test_mysql_database_engine/test.py::test_mysql_ddl_for_mysql_database test_parallel_replicas_skip_shards/test.py::test_skip_unavailable_shards [gw0] [ 50%] FAILED test_parallel_replicas_skip_shards/test.py::test_skip_unavailable_shards [gw1] [100%] PASSED test_mysql_database_engine/test.py::test_mysql_ddl_for_mysql_database =================================== FAILURES =================================== _________________________ test_skip_unavailable_shards _________________________ [gw0] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_skip_unavailable_shards(start_cluster): expected = "node1\nnode2\nnode3\n" > assert ( node1.query( "SELECT hostName() as h FROM clusterAllReplicas('two_shards', system.one) order by h", settings={ "allow_experimental_parallel_reading_from_replicas": 0, "skip_unavailable_shards": 1, }, ) == expected ) E AssertionError: assert 'node1\nnode1\nnode1\n' == 'node1\nnode2\nnode3\n' E node1 E - node2 E ? ^ E + node1 E ? ^ E - node3 E ? ^ E + node1 E ? ^ test_parallel_replicas_skip_shards/test.py:24: AssertionError ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ---------------------------- Captured stderr setup ----------------------------- WARNING: API is accessible on http://0.0.0.0:2375 without encryption. Access to the remote API is equivalent to root access on the host. Refer to the 'Docker daemon attack surface' section in the documentation for more information: https://docs.docker.com/go/attack-surface/ WARNING: API is accessible on http://0.0.0.0:2375 without encryption. Access to the remote API is equivalent to root access on the host. Refer to the 'Docker daemon attack surface' section in the documentation for more information: https://docs.docker.com/go/attack-surface/ ------------------------------ Captured log setup ------------------------------ 2024-09-24 00:19:16 [ 368 ] DEBUG : Command:['docker ps | wc -l'] (cluster.py:105, run_and_check) 2024-09-24 00:19:16 [ 368 ] DEBUG : Stdout:1 (cluster.py:113, run_and_check) 2024-09-24 00:19:16 [ 368 ] DEBUG : No running containers (conftest.py:60, cleanup_environment) 2024-09-24 00:19:16 [ 368 ] DEBUG : Pruning Docker networks (conftest.py:62, cleanup_environment) 2024-09-24 00:19:16 [ 368 ] DEBUG : Command:['docker network prune --force'] (cluster.py:105, run_and_check) 2024-09-24 00:19:16 [ 368 ] DEBUG : Command:["sysctl net.ipv4.ip_local_port_range='55000 65535'"] (cluster.py:105, run_and_check) 2024-09-24 00:19:16 [ 368 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:113, run_and_check) 2024-09-24 00:19:16 [ 368 ] INFO : Running tests in /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/test.py (cluster.py:2637, start) 2024-09-24 00:19:16 [ 368 ] DEBUG : Cluster start called. is_up=False (cluster.py:2644, start) 2024-09-24 00:19:16 [ 368 ] DEBUG : !!! Docker info: Client: Docker Engine - Community Version: 27.2.1 Context: default Debug Mode: false Plugins: buildx: Docker Buildx (Docker Inc.) Version: v0.16.2 Path: /usr/libexec/docker/cli-plugins/docker-buildx compose: Docker Compose (Docker Inc.) Version: v2.29.2 Path: /usr/libexec/docker/cli-plugins/docker-compose Server: Containers: 0 Running: 0 Paused: 0 Stopped: 0 Images: 45 Server Version: 23.0.6 Storage Driver: overlay2 Backing Filesystem: extfs Supports d_type: true Using metacopy: false Native Overlay Diff: true userxattr: false Logging Driver: json-file Cgroup Driver: cgroupfs Cgroup Version: 2 Plugins: Volume: local Network: bridge host ipvlan macvlan null overlay Log: awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog Swarm: inactive Runtimes: io.containerd.runc.v2 runc Default Runtime: runc Init Binary: docker-init containerd version: 7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c runc version: v1.1.14-0-g2c9f560 init version: de40ad0 Security Options: seccomp Profile: builtin cgroupns Kernel Version: 5.15.0-119-generic Operating System: Ubuntu 22.04.5 LTS (containerized) OSType: linux Architecture: x86_64 CPUs: 16 Total Memory: 30.6GiB Name: 6a011b450f4c ID: f3cee668-1847-47db-b7b0-0381d2ad5f93 Docker Root Dir: /var/lib/docker Debug Mode: false Experimental: false Insecure Registries: 65.108.242.32:5000 127.0.0.0/8 Registry Mirrors: http://65.108.242.32:5000/ Live Restore Enabled: false Default Address Pools: Base: 172.17.0.0/12, Size: 24 (cluster.py:722, print_all_docker_pieces) 2024-09-24 00:19:16 [ 368 ] DEBUG : Docker networks for project roottestparallelreplicasskipshards are NETWORK ID NAME DRIVER SCOPE (cluster.py:733, print_all_docker_pieces) 2024-09-24 00:19:16 [ 368 ] DEBUG : Docker containers for project roottestparallelreplicasskipshards are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:741, print_all_docker_pieces) 2024-09-24 00:19:16 [ 368 ] DEBUG : Docker volumes for project roottestparallelreplicasskipshards are DRIVER VOLUME NAME (cluster.py:749, print_all_docker_pieces) 2024-09-24 00:19:16 [ 368 ] DEBUG : Cleanup called (cluster.py:754, cleanup) 2024-09-24 00:19:16 [ 368 ] DEBUG : !!! Docker info: Client: Docker Engine - Community Version: 27.2.1 Context: default Debug Mode: false Plugins: buildx: Docker Buildx (Docker Inc.) Version: v0.16.2 Path: /usr/libexec/docker/cli-plugins/docker-buildx compose: Docker Compose (Docker Inc.) Version: v2.29.2 Path: /usr/libexec/docker/cli-plugins/docker-compose Server: Containers: 0 Running: 0 Paused: 0 Stopped: 0 Images: 45 Server Version: 23.0.6 Storage Driver: overlay2 Backing Filesystem: extfs Supports d_type: true Using metacopy: false Native Overlay Diff: true userxattr: false Logging Driver: json-file Cgroup Driver: cgroupfs Cgroup Version: 2 Plugins: Volume: local Network: bridge host ipvlan macvlan null overlay Log: awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog Swarm: inactive Runtimes: io.containerd.runc.v2 runc Default Runtime: runc Init Binary: docker-init containerd version: 7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c runc version: v1.1.14-0-g2c9f560 init version: de40ad0 Security Options: seccomp Profile: builtin cgroupns Kernel Version: 5.15.0-119-generic Operating System: Ubuntu 22.04.5 LTS (containerized) OSType: linux Architecture: x86_64 CPUs: 16 Total Memory: 30.6GiB Name: 6a011b450f4c ID: f3cee668-1847-47db-b7b0-0381d2ad5f93 Docker Root Dir: /var/lib/docker Debug Mode: false Experimental: false Insecure Registries: 65.108.242.32:5000 127.0.0.0/8 Registry Mirrors: http://65.108.242.32:5000/ Live Restore Enabled: false Default Address Pools: Base: 172.17.0.0/12, Size: 24 (cluster.py:722, print_all_docker_pieces) 2024-09-24 00:19:17 [ 368 ] DEBUG : Docker networks for project roottestparallelreplicasskipshards are NETWORK ID NAME DRIVER SCOPE (cluster.py:733, print_all_docker_pieces) 2024-09-24 00:19:17 [ 368 ] DEBUG : Docker containers for project roottestparallelreplicasskipshards are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:741, print_all_docker_pieces) 2024-09-24 00:19:17 [ 368 ] DEBUG : Docker volumes for project roottestparallelreplicasskipshards are DRIVER VOLUME NAME (cluster.py:749, print_all_docker_pieces) 2024-09-24 00:19:17 [ 368 ] DEBUG : Command:docker container list --all --filter name='^/roottestparallelreplicasskipshards_.*_1$' --format '{{.ID}}:{{.Names}}' (cluster.py:105, run_and_check) 2024-09-24 00:19:17 [ 368 ] DEBUG : Unstopped containers: {} (cluster.py:768, cleanup) 2024-09-24 00:19:17 [ 368 ] DEBUG : No running containers for project: roottestparallelreplicasskipshards (cluster.py:782, cleanup) 2024-09-24 00:19:17 [ 368 ] DEBUG : Trying to prune unused networks... (cluster.py:788, cleanup) 2024-09-24 00:19:17 [ 368 ] DEBUG : Trying to prune unused images... (cluster.py:804, cleanup) 2024-09-24 00:19:17 [ 368 ] DEBUG : Command:['docker', 'image', 'prune', '-f'] (cluster.py:105, run_and_check) 2024-09-24 00:19:17 [ 368 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:113, run_and_check) 2024-09-24 00:19:17 [ 368 ] DEBUG : Images pruned (cluster.py:807, cleanup) 2024-09-24 00:19:17 [ 368 ] DEBUG : Trying to prune unused volumes... (cluster.py:813, cleanup) 2024-09-24 00:19:17 [ 368 ] DEBUG : Command:['docker volume ls | wc -l'] (cluster.py:105, run_and_check) 2024-09-24 00:19:17 [ 368 ] DEBUG : Stdout:1 (cluster.py:113, run_and_check) 2024-09-24 00:19:17 [ 368 ] DEBUG : Setup directory for instance: node1 (cluster.py:2657, start) 2024-09-24 00:19:17 [ 368 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4309, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Create directory for common tests configuration (cluster.py:4314, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Copy common configuration from helpers (cluster.py:4334, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Generate and write macros file (cluster.py:4353, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/configs/config.d (cluster.py:4383, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/database (cluster.py:4400, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/logs (cluster.py:4411, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log"] (cluster.py:4487, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Setup directory for instance: node2 (cluster.py:2657, start) 2024-09-24 00:19:17 [ 368 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4309, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Create directory for common tests configuration (cluster.py:4314, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Copy common configuration from helpers (cluster.py:4334, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Generate and write macros file (cluster.py:4353, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/configs/config.d (cluster.py:4383, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/database (cluster.py:4400, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/logs (cluster.py:4411, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log"] (cluster.py:4487, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Setup directory for instance: node3 (cluster.py:2657, start) 2024-09-24 00:19:17 [ 368 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4309, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Create directory for common tests configuration (cluster.py:4314, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Copy common configuration from helpers (cluster.py:4334, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Generate and write macros file (cluster.py:4353, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/configs/config.d (cluster.py:4383, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/database (cluster.py:4400, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/logs (cluster.py:4411, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log"] (cluster.py:4487, create_dir) 2024-09-24 00:19:17 [ 368 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw'} stored in /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/.env (cluster.py:78, _create_env_file) 2024-09-24 00:19:17 [ 368 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2024-09-24 00:19:17 [ 368 ] DEBUG : No config file found (config.py:28, find_config_file) 2024-09-24 00:19:17 [ 368 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2024-09-24 00:19:17 [ 368 ] DEBUG : No config file found (config.py:28, find_config_file) 2024-09-24 00:19:17 [ 368 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:546, _make_request) 2024-09-24 00:19:17 [ 368 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/.env', '--project-name', 'roottestparallelreplicasskipshards', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/docker-compose.yml', 'pull'] (cluster.py:105, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node2 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node3 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node1 ... pulling from altinityinfra/integr... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node2 ... pulling from altinityinfra/integr... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node1 ... digest: sha256:3c12bcb41068a0596b... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node1 ... status: image is up to date for a... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node2 ... digest: sha256:3c12bcb41068a0596b... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node2 ... status: image is up to date for a... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node2 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node3 ... pulling from altinityinfra/integr... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node3 ... digest: sha256:3c12bcb41068a0596b... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node3 ... status: image is up to date for a... (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : Stderr:Pulling node3 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:28 [ 368 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker-compose --env-file /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/.env --project-name roottestparallelreplicasskipshards --file /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/docker-compose.yml --file /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/docker-compose.yml --file /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/docker-compose.yml up -d --no-recreate') (cluster.py:2980, start) 2024-09-24 00:19:28 [ 368 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/.env', '--project-name', 'roottestparallelreplicasskipshards', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/docker-compose.yml', 'up', '-d', '--no-recreate'] (cluster.py:105, run_and_check) 2024-09-24 00:19:29 [ 368 ] DEBUG : Stderr:Creating network "roottestparallelreplicasskipshards_default" with the default driver (cluster.py:115, run_and_check) 2024-09-24 00:19:29 [ 368 ] DEBUG : Stderr:Creating roottestparallelreplicasskipshards_node3_1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:29 [ 368 ] DEBUG : Stderr:Creating roottestparallelreplicasskipshards_node1_1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:29 [ 368 ] DEBUG : Stderr:Creating roottestparallelreplicasskipshards_node2_1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:29 [ 368 ] DEBUG : Stderr:Creating roottestparallelreplicasskipshards_node3_1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:29 [ 368 ] DEBUG : Stderr:Creating roottestparallelreplicasskipshards_node1_1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:29 [ 368 ] DEBUG : Stderr:Creating roottestparallelreplicasskipshards_node2_1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:29 [ 368 ] DEBUG : ClickHouse instance created (cluster.py:2988, start) 2024-09-24 00:19:29 [ 368 ] DEBUG : get_instance_ip instance_name=node1 (cluster.py:1971, get_instance_ip) 2024-09-24 00:19:29 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/roottestparallelreplicasskipshards_node1_1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:29 [ 368 ] DEBUG : Waiting for ClickHouse start in node1, ip: 172.16.2.3... (cluster.py:2995, start) 2024-09-24 00:19:29 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/roottestparallelreplicasskipshards_node1_1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:29 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:29 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:29 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:30 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:30 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:30 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:30 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:30 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:30 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:30 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:30 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:30 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/134276b0470e24208de36c0880c128ea6f8e7d7a0c12e413ffd55fc5bc3a5ec8/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : ClickHouse node1 started (cluster.py:2999, start) 2024-09-24 00:19:31 [ 368 ] DEBUG : get_instance_ip instance_name=node2 (cluster.py:1971, get_instance_ip) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/roottestparallelreplicasskipshards_node2_1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : Waiting for ClickHouse start in node2, ip: 172.16.2.4... (cluster.py:2995, start) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/roottestparallelreplicasskipshards_node2_1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/12346f45b2082ae794d866e50a1a075b2a32a62ae956f346e160ddf1e2c1e47f/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/12346f45b2082ae794d866e50a1a075b2a32a62ae956f346e160ddf1e2c1e47f/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : ClickHouse node2 started (cluster.py:2999, start) 2024-09-24 00:19:31 [ 368 ] DEBUG : get_instance_ip instance_name=node3 (cluster.py:1971, get_instance_ip) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/roottestparallelreplicasskipshards_node3_1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : Waiting for ClickHouse start in node3, ip: 172.16.2.2... (cluster.py:2995, start) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/roottestparallelreplicasskipshards_node3_1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : http://localhost:None "GET /v1.42/containers/c0d5fe5f65bbff6ee3c5851016e8aecaaa182b250654bcdfc683b5cc21447fb1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2024-09-24 00:19:31 [ 368 ] DEBUG : ClickHouse node3 started (cluster.py:2999, start) ------------------------------ Captured log call ------------------------------- 2024-09-24 00:19:31 [ 368 ] DEBUG : Executing query SELECT hostName() as h FROM clusterAllReplicas('two_shards', system.one) order by h on node1 (cluster.py:3423, query) --------------------------- Captured stderr teardown --------------------------- WARNING: API is accessible on http://0.0.0.0:2375 without encryption. Access to the remote API is equivalent to root access on the host. Refer to the 'Docker daemon attack surface' section in the documentation for more information: https://docs.docker.com/go/attack-surface/ ---------------------------- Captured log teardown ----------------------------- 2024-09-24 00:19:32 [ 368 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/.env', '--project-name', 'roottestparallelreplicasskipshards', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/docker-compose.yml', 'stop', '--timeout', '20'] (cluster.py:105, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Stderr:Stopping roottestparallelreplicasskipshards_node2_1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Stderr:Stopping roottestparallelreplicasskipshards_node1_1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Stderr:Stopping roottestparallelreplicasskipshards_node3_1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Stderr:Stopping roottestparallelreplicasskipshards_node1_1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Stderr:Stopping roottestparallelreplicasskipshards_node3_1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Stderr:Stopping roottestparallelreplicasskipshards_node2_1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Command:['bash', '-c', '[ -f /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/logs/stderr.log* || true'] (cluster.py:105, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Command:['bash', '-c', '[ -f /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/logs/stderr.log* || true'] (cluster.py:105, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Command:['bash', '-c', '[ -f /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/logs/stderr.log* || true'] (cluster.py:105, run_and_check) 2024-09-24 00:19:35 [ 368 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/.env', '--project-name', 'roottestparallelreplicasskipshards', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node1/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node2/docker-compose.yml', '--file', '/ClickHouse/tests/integration/test_parallel_replicas_skip_shards/_instances_2/node3/docker-compose.yml', 'down', '--volumes'] (cluster.py:105, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Stderr:Removing roottestparallelreplicasskipshards_node2_1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Stderr:Removing roottestparallelreplicasskipshards_node1_1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Stderr:Removing roottestparallelreplicasskipshards_node3_1 ... (cluster.py:115, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Stderr:Removing roottestparallelreplicasskipshards_node3_1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Stderr:Removing roottestparallelreplicasskipshards_node2_1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Stderr:Removing roottestparallelreplicasskipshards_node1_1 ... done (cluster.py:115, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Stderr:Removing network roottestparallelreplicasskipshards_default (cluster.py:115, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Cleanup called (cluster.py:754, cleanup) 2024-09-24 00:19:36 [ 368 ] DEBUG : !!! Docker info: Client: Docker Engine - Community Version: 27.2.1 Context: default Debug Mode: false Plugins: buildx: Docker Buildx (Docker Inc.) Version: v0.16.2 Path: /usr/libexec/docker/cli-plugins/docker-buildx compose: Docker Compose (Docker Inc.) Version: v2.29.2 Path: /usr/libexec/docker/cli-plugins/docker-compose Server: Containers: 1 Running: 1 Paused: 0 Stopped: 0 Images: 45 Server Version: 23.0.6 Storage Driver: overlay2 Backing Filesystem: extfs Supports d_type: true Using metacopy: false Native Overlay Diff: true userxattr: false Logging Driver: json-file Cgroup Driver: cgroupfs Cgroup Version: 2 Plugins: Volume: local Network: bridge host ipvlan macvlan null overlay Log: awslogs fluentd gcplogs gelf journald json-file local logentries splunk syslog Swarm: inactive Runtimes: io.containerd.runc.v2 runc Default Runtime: runc Init Binary: docker-init containerd version: 7f7fdf5fed64eb6a7caf99b3e12efcf9d60e311c runc version: v1.1.14-0-g2c9f560 init version: de40ad0 Security Options: seccomp Profile: builtin cgroupns Kernel Version: 5.15.0-119-generic Operating System: Ubuntu 22.04.5 LTS (containerized) OSType: linux Architecture: x86_64 CPUs: 16 Total Memory: 30.6GiB Name: 6a011b450f4c ID: f3cee668-1847-47db-b7b0-0381d2ad5f93 Docker Root Dir: /var/lib/docker Debug Mode: false Experimental: false Insecure Registries: 65.108.242.32:5000 127.0.0.0/8 Registry Mirrors: http://65.108.242.32:5000/ Live Restore Enabled: false Default Address Pools: Base: 172.17.0.0/12, Size: 24 (cluster.py:722, print_all_docker_pieces) 2024-09-24 00:19:36 [ 368 ] DEBUG : Docker networks for project roottestparallelreplicasskipshards are NETWORK ID NAME DRIVER SCOPE (cluster.py:733, print_all_docker_pieces) 2024-09-24 00:19:36 [ 368 ] DEBUG : Docker containers for project roottestparallelreplicasskipshards are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:741, print_all_docker_pieces) 2024-09-24 00:19:36 [ 368 ] DEBUG : Docker volumes for project roottestparallelreplicasskipshards are DRIVER VOLUME NAME (cluster.py:749, print_all_docker_pieces) 2024-09-24 00:19:36 [ 368 ] DEBUG : Command:docker container list --all --filter name='^/roottestparallelreplicasskipshards_.*_1$' --format '{{.ID}}:{{.Names}}' (cluster.py:105, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Unstopped containers: {} (cluster.py:768, cleanup) 2024-09-24 00:19:36 [ 368 ] DEBUG : No running containers for project: roottestparallelreplicasskipshards (cluster.py:782, cleanup) 2024-09-24 00:19:36 [ 368 ] DEBUG : Trying to prune unused networks... (cluster.py:788, cleanup) 2024-09-24 00:19:36 [ 368 ] DEBUG : Trying to prune unused images... (cluster.py:804, cleanup) 2024-09-24 00:19:36 [ 368 ] DEBUG : Command:['docker', 'image', 'prune', '-f'] (cluster.py:105, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:113, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Images pruned (cluster.py:807, cleanup) 2024-09-24 00:19:36 [ 368 ] DEBUG : Trying to prune unused volumes... (cluster.py:813, cleanup) 2024-09-24 00:19:36 [ 368 ] DEBUG : Command:['docker volume ls | wc -l'] (cluster.py:105, run_and_check) 2024-09-24 00:19:36 [ 368 ] DEBUG : Stdout:2 (cluster.py:113, run_and_check) ============================== slowest durations =============================== 22.75s setup test_mysql_database_engine/test.py::test_mysql_ddl_for_mysql_database 15.16s setup test_parallel_replicas_skip_shards/test.py::test_skip_unavailable_shards 7.84s call test_mysql_database_engine/test.py::test_mysql_ddl_for_mysql_database 4.43s teardown test_parallel_replicas_skip_shards/test.py::test_skip_unavailable_shards 3.96s teardown test_mysql_database_engine/test.py::test_mysql_ddl_for_mysql_database 0.27s call test_parallel_replicas_skip_shards/test.py::test_skip_unavailable_shards =========================== short test summary info ============================ FAILED test_parallel_replicas_skip_shards/test.py::test_skip_unavailable_shards PASSED test_mysql_database_engine/test.py::test_mysql_ddl_for_mysql_database ========================= 1 failed, 1 passed in 35.56s ========================= Traceback (most recent call last): File "/home/ubuntu/_work/_temp/test/git-repo-copy/tests/integration/./runner", line 455, in subprocess.check_call(cmd, shell=True) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_o9qf3p --privileged --dns-search='.' --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/_temp/test/git-repo-copy/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/_temp/test/git-repo-copy/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/_temp/test/git-repo-copy/src/Server/grpc_protos:/ClickHouse/src/Server/grpc_protos --volume=/run:/run/host:ro --mount type=bind,source=/home/ubuntu/_work/_temp/test/dockerd_volume_dir,target=/var/lib/docker -e DOCKER_HELPER_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_BASE_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_KERBERIZED_HADOOP_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_KERBEROS_KDC_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_MYSQL_JAVA_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_MYSQL_JS_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_MYSQL_PHP_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_NGINX_DAV_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e CLICKHOUSE_USE_NEW_ANALYZER=1 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 2 -rfEps --run-id=2 --color=no --durations=0 test_mysql_database_engine/test.py::test_mysql_ddl_for_mysql_database test_parallel_replicas_skip_shards/test.py::test_skip_unavailable_shards -vvv" altinityinfra/integration-tests-runner:0-9992f5eba3591142ff413d31cb83a7bf4f0e2553 ' returned non-zero exit status 1.